Adaptive Perturbation-Based Gradient Estimation for Discrete Latent Variable Models

نویسندگان

چکیده

The integration of discrete algorithmic components in deep learning architectures has numerous applications. Recently, Implicit Maximum Likelihood Estimation, a class gradient estimators for exponential family distributions, was proposed by combining implicit differentiation through perturbation with the path-wise estimator. However, due to finite difference approximation gradients, it is especially sensitive choice step size, which needs be specified user. In this work, we present Adaptive IMLE (AIMLE), first adaptive estimator complex distributions: adaptively identifies target distribution trading off density information degree bias estimates. We empirically evaluate our on synthetic examples, as well Learning Explain, Discrete Variational Auto-Encoders, and Neural Relational Inference tasks. experiments, show that can produce faithful estimates while requiring orders magnitude fewer samples than other estimators.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

REBAR: Low-variance, unbiased gradient estimates for discrete latent variable models

Learning in models with discrete latent variables is challenging due to high variance gradient estimators. Generally, approaches have relied on control variates to reduce the variance of the REINFORCE estimator. Recent work (Jang et al., 2016; Maddison et al., 2016) has taken a different approach, introducing a continuous relaxation of discrete variables to produce low-variance, but biased, gra...

متن کامل

Simulation-based approach to estimation of latent variable models

We propose a simulation-based method for calculating maximum likelihood estimators in latent variable models. The proposed method integrates a recently developed sampling strategy, the so-called Sample Average Approximation method, to efficiently compute high quality solutions of the estimation problem. Theoretical and algorithmic properties of the method are discussed. A computational study, i...

متن کامل

Estimation of Generalized Linear Latent Variable Models

Generalized Linear Latent Variable Models (GLLVM), as de ned in Bartholomew and Knott (1999) enable modelling of relationships between manifest and latent variables. They extend structural equation modelling techniques, which are powerful tools in the social sciences. However, because of the complexity of the log-likelihood function of a GLLVM, an approximation such as numerical integration mus...

متن کامل

Nonparametric Bayesian discrete latent variable models for unsupervised learning

The analysis of real-world problems often requires robust and flexible models that can accurately represent the structure in the data. Nonparametric Bayesian priors allow the construction of such models which can be used for complex real-world data. Nonparametric models, despite their name, can be defined as models that have infinitely many parameters. This thesis is about two types of nonparam...

متن کامل

Bayesian latent variable models for mixed discrete outcomes.

In studies of complex health conditions, mixtures of discrete outcomes (event time, count, binary, ordered categorical) are commonly collected. For example, studies of skin tumorigenesis record latency time prior to the first tumor, increases in the number of tumors at each week, and the occurrence of internal tumors at the time of death. Motivated by this application, we propose a general unde...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2023

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v37i8.26103